Decentralized Learning With Lazy and Approximate Dual Gradients

نویسندگان

چکیده

This paper develops algorithms for decentralized machine learning over a network, where data are distributed, computation is localized, and communication restricted between neighbors. A line of recent research in this area focuses on improving both complexities. The methods SSDA MSDA (Scaman et al., 2017) have optimal complexity when the objective smooth strongly convex, simple to derive. However, they require solving subproblem at each step, so required accuracy solutions total computational complexities uncertain. We propose new that instead subproblem, run warm-started Katyusha small, fixed number steps. In addition, previous information sufficiently useful, local rule will decide even skip round communication, leading extra savings. show our efficient provably reducing MSDA. numerical experiments, achieve significant reduction compared with state-of-the-art.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Dual Control for Approximate Bayesian Reinforcement Learning

Control of non-episodic, finite-horizon dynamical systems with uncertain dynamics poses a tough and elementary case of the exploration-exploitation trade-off. Bayesian reinforcement learning, reasoning about the effect of actions and future observations, offers a principled solution, but is intractable. We review, then extend an old approximate approach from control theory—where the problem is ...

متن کامل

Learning Efficiently with Approximate Inference via Dual Losses

Many structured prediction tasks involve complex models where inference is computationally intractable, but where it can be well approximated using a linear programming relaxation. Previous approaches for learning for structured prediction (e.g., cuttingplane, subgradient methods, perceptron) repeatedly make predictions for some of the data points. These approaches are computationally demanding...

متن کامل

Approximate distance fields with non-vanishing gradients

For a given set of points S, a Euclidean distance field is defined by associating with every point p of Euclidean space Ed a value that is equal to the Euclidean distance from p to S. Such distance fields have numerous computational applications, but are expensive to compute and may not be sufficiently smooth for some applications. Instead, popular implicit modeling techniques rely on various a...

متن کامل

Approximate Decentralized Bayesian Inference

This paper presents an approximate method for performing Bayesian inference in models with conditional independence over a decentralized network of learning agents. The method first employs variational inference on each individual learning agent to generate a local approximate posterior, the agents transmit their local posteriors to other agents in the network, and finally each agent combines i...

متن کامل

Decentralized Consensus Algorithm with Delayed and Stochastic Gradients

Abstract. We analyze the convergence of a decentralized consensus algorithm with delayed gradient information across the network. The nodes in the network privately hold parts of the objective function and collaboratively solve for the consensus optimal solution of the total objective while they can only communicate with their immediate neighbors. In real-world networks, it is often difficult a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Signal Processing

سال: 2021

ISSN: ['1053-587X', '1941-0476']

DOI: https://doi.org/10.1109/tsp.2021.3056915